-
Notifications
You must be signed in to change notification settings - Fork 15.2k
[DAGCombiner] Spill dynamic insertelt chain in one go #162368
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
@llvm/pr-subscribers-backend-nvptx Author: Princeton Ferro (Prince781) ChangesA chain of dynamic insertelts (that is: insertelt (insertelt (...)) with dynamic indices) can be spilled at once. This avoids each insertelt being spilled in DAGTypeLegalizer which reduces code size and compile time. Full diff: https://github.com/llvm/llvm-project/pull/162368.diff 2 Files Affected:
diff --git a/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp b/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
index 309f1bea8b77c..3071aac3a511a 100644
--- a/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
+++ b/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
@@ -23445,6 +23445,57 @@ SDValue DAGCombiner::visitINSERT_VECTOR_ELT(SDNode *N) {
// inselt undef, InVal, EltNo --> build_vector < InVal, InVal, ... >
if (InVec.isUndef() && TLI.shouldSplatInsEltVarIndex(VT))
return DAG.getSplat(VT, DL, InVal);
+
+ // For dynamic insertelts, the type legalizer will spill the entire vector.
+ // For a chain of dynamic insertelts, this can be really inefficient and
+ // bad for compile time. If each insertelt is only fed into the next, the
+ // vector is write-only across this chain, and we can just spill once.
+ SmallVector<SDNode *> Seq{N};
+ while (true) {
+ SDValue InVec = Seq.back()->getOperand(0);
+ SDValue EltNo = InVec.getOperand(2);
+ if (!(InVec.getOpcode() == ISD::INSERT_VECTOR_ELT &&
+ !isa<ConstantSDNode>(EltNo)))
+ break;
+ Seq.push_back(InVec.getNode());
+ }
+
+ // Only care about chains, otherwise this instruction can be handled by
+ // the type legalizer just fine.
+ if (Seq.size() > 1) {
+ // In cases where the vector is illegal it will be broken down into parts
+ // and stored in parts - we should use the alignment for the smallest part.
+ Align SmallestAlign = DAG.getReducedAlign(VT, /*UseABI=*/false);
+ SDValue StackPtr = DAG.CreateStackTemporary(VT.getStoreSize(), SmallestAlign);
+ auto &MF = DAG.getMachineFunction();
+ auto FrameIndex = cast<FrameIndexSDNode>(StackPtr.getNode())->getIndex();
+ auto PtrInfo = MachinePointerInfo::getFixedStack(MF, FrameIndex);
+
+ // Begin spilling
+ SDValue InVec = Seq.back()->getOperand(0);
+ SDValue Store = DAG.getStore(DAG.getEntryNode(), DL, InVec, StackPtr, PtrInfo,
+ SmallestAlign);
+
+ // Lower each dynamic insertelt to a store
+ for (SDNode *N : reverse(Seq)) {
+ SDValue Elmnt = N->getOperand(1);
+ SDValue Index = N->getOperand(2);
+
+ // Store the new element. This may be larger than the vector element type,
+ // so use a truncating store.
+ SDValue EltPtr = TLI.getVectorElementPointer(DAG, StackPtr, VT, Index);
+ EVT EltVT = Elmnt.getValueType();
+ Store = DAG.getTruncStore(
+ Store, DL, Elmnt, EltPtr, MachinePointerInfo::getUnknownStack(MF),
+ EltVT, commonAlignment(SmallestAlign,
+ EltVT.getFixedSizeInBits() / 8));
+ }
+
+ // Load the spilled vector
+ SDValue Load = DAG.getLoad(VT, DL, Store, StackPtr, PtrInfo, SmallestAlign);
+ return Load.getValue(0);
+ }
+
return SDValue();
}
diff --git a/llvm/test/CodeGen/NVPTX/vector-spill.ll b/llvm/test/CodeGen/NVPTX/vector-spill.ll
new file mode 100644
index 0000000000000..0a26069b02cc2
--- /dev/null
+++ b/llvm/test/CodeGen/NVPTX/vector-spill.ll
@@ -0,0 +1,215 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 6
+; RUN: llc < %s -mcpu=sm_20 | FileCheck %s
+; RUN: %if ptxas %{ llc < %s -mcpu=sm_20 | %ptxas-verify %}
+target datalayout = "e-i64:64-v16:16-v32:32-n16:32:64"
+target triple = "nvptx64-nvidia-cuda"
+
+; COM: Spill the vector once.
+define ptx_kernel void @spill_once(ptr addrspace(3) %shared.mem, <8 x double> %vector, i32 %idx0, i32 %idx1, i32 %idx2, i32 %idx3) local_unnamed_addr {
+; CHECK-LABEL: spill_once(
+; CHECK: {
+; CHECK-NEXT: .local .align 8 .b8 __local_depot0[64];
+; CHECK-NEXT: .reg .b64 %SP;
+; CHECK-NEXT: .reg .b64 %SPL;
+; CHECK-NEXT: .reg .b64 %rd<39>;
+; CHECK-EMPTY:
+; CHECK-NEXT: // %bb.0: // %entry
+; CHECK-NEXT: mov.b64 %SPL, __local_depot0;
+; CHECK-NEXT: cvta.local.u64 %SP, %SPL;
+; CHECK-NEXT: ld.param.b64 %rd1, [spill_once_param_0];
+; CHECK-NEXT: ld.param.v2.b64 {%rd2, %rd3}, [spill_once_param_1];
+; CHECK-NEXT: ld.param.v2.b64 {%rd4, %rd5}, [spill_once_param_1+16];
+; CHECK-NEXT: ld.param.v2.b64 {%rd6, %rd7}, [spill_once_param_1+32];
+; CHECK-NEXT: ld.param.v2.b64 {%rd8, %rd9}, [spill_once_param_1+48];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd10, %rd11}, [%rd1];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd12, %rd13}, [%rd1+16];
+; CHECK-NEXT: ld.param.b32 %rd14, [spill_once_param_2];
+; CHECK-NEXT: shl.b64 %rd15, %rd14, 3;
+; CHECK-NEXT: and.b64 %rd16, %rd15, 56;
+; CHECK-NEXT: add.u64 %rd17, %SP, 0;
+; CHECK-NEXT: add.s64 %rd18, %rd17, %rd16;
+; CHECK-NEXT: ld.param.b32 %rd19, [spill_once_param_3];
+; CHECK-NEXT: shl.b64 %rd20, %rd19, 3;
+; CHECK-NEXT: and.b64 %rd21, %rd20, 56;
+; CHECK-NEXT: add.s64 %rd22, %rd17, %rd21;
+; CHECK-NEXT: ld.param.b32 %rd23, [spill_once_param_4];
+; CHECK-NEXT: shl.b64 %rd24, %rd23, 3;
+; CHECK-NEXT: and.b64 %rd25, %rd24, 56;
+; CHECK-NEXT: add.s64 %rd26, %rd17, %rd25;
+; CHECK-NEXT: st.b64 [%SP+56], %rd9;
+; CHECK-NEXT: st.b64 [%SP+48], %rd8;
+; CHECK-NEXT: st.b64 [%SP+40], %rd7;
+; CHECK-NEXT: st.b64 [%SP+32], %rd6;
+; CHECK-NEXT: st.b64 [%SP+24], %rd5;
+; CHECK-NEXT: st.b64 [%SP+16], %rd4;
+; CHECK-NEXT: st.b64 [%SP+8], %rd3;
+; CHECK-NEXT: st.b64 [%SP], %rd2;
+; CHECK-NEXT: st.b64 [%rd18], %rd10;
+; CHECK-NEXT: st.b64 [%rd22], %rd11;
+; CHECK-NEXT: st.b64 [%rd26], %rd12;
+; CHECK-NEXT: ld.param.b32 %rd27, [spill_once_param_5];
+; CHECK-NEXT: shl.b64 %rd28, %rd27, 3;
+; CHECK-NEXT: and.b64 %rd29, %rd28, 56;
+; CHECK-NEXT: add.s64 %rd30, %rd17, %rd29;
+; CHECK-NEXT: st.b64 [%rd30], %rd13;
+; CHECK-NEXT: ld.b64 %rd31, [%SP+8];
+; CHECK-NEXT: ld.b64 %rd32, [%SP];
+; CHECK-NEXT: ld.b64 %rd33, [%SP+24];
+; CHECK-NEXT: ld.b64 %rd34, [%SP+16];
+; CHECK-NEXT: ld.b64 %rd35, [%SP+40];
+; CHECK-NEXT: ld.b64 %rd36, [%SP+32];
+; CHECK-NEXT: ld.b64 %rd37, [%SP+56];
+; CHECK-NEXT: ld.b64 %rd38, [%SP+48];
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1072], {%rd38, %rd37};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1056], {%rd36, %rd35};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1040], {%rd34, %rd33};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1024], {%rd32, %rd31};
+; CHECK-NEXT: ret;
+entry:
+ %offset.0 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 0
+ %element.0 = load double, ptr addrspace(3) %offset.0, align 64
+ %offset.1 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 8
+ %element.1 = load double, ptr addrspace(3) %offset.1, align 8
+ %offset.2 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 16
+ %element.2 = load double, ptr addrspace(3) %offset.2, align 8
+ %offset.3 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 24
+ %element.3 = load double, ptr addrspace(3) %offset.3, align 8
+ %vector.build0 = insertelement <8 x double> %vector, double %element.0, i32 %idx0
+ %vector.build1 = insertelement <8 x double> %vector.build0, double %element.1, i32 %idx1
+ %vector.build2 = insertelement <8 x double> %vector.build1, double %element.2, i32 %idx2
+ %vector.build3 = insertelement <8 x double> %vector.build2, double %element.3, i32 %idx3
+ %location = getelementptr i8, ptr addrspace(3) %shared.mem, i32 1024
+ store <8 x double> %vector.build3, ptr addrspace(3) %location, align 64
+ ret void
+}
+
+; COM: Spill the vector twice. Because these are in two different slots, the
+; resulting spill codes may be non-overlapping even though the insertelt
+; sequences overlap.
+define ptx_kernel void @spill_twice(ptr addrspace(3) %shared.mem, <8 x double> %vector, i32 %idx0, i32 %idx1, i32 %idx2, i32 %idx3) local_unnamed_addr {
+; CHECK-LABEL: spill_twice(
+; CHECK: {
+; CHECK-NEXT: .local .align 8 .b8 __local_depot1[128];
+; CHECK-NEXT: .reg .b64 %SP;
+; CHECK-NEXT: .reg .b64 %SPL;
+; CHECK-NEXT: .reg .b32 %r<7>;
+; CHECK-NEXT: .reg .b64 %rd<54>;
+; CHECK-EMPTY:
+; CHECK-NEXT: // %bb.0: // %entry
+; CHECK-NEXT: mov.b64 %SPL, __local_depot1;
+; CHECK-NEXT: cvta.local.u64 %SP, %SPL;
+; CHECK-NEXT: ld.param.b64 %rd1, [spill_twice_param_0];
+; CHECK-NEXT: ld.param.v2.b64 {%rd2, %rd3}, [spill_twice_param_1];
+; CHECK-NEXT: ld.param.v2.b64 {%rd4, %rd5}, [spill_twice_param_1+16];
+; CHECK-NEXT: ld.param.v2.b64 {%rd6, %rd7}, [spill_twice_param_1+32];
+; CHECK-NEXT: ld.param.v2.b64 {%rd8, %rd9}, [spill_twice_param_1+48];
+; CHECK-NEXT: ld.param.b32 %r1, [spill_twice_param_2];
+; CHECK-NEXT: ld.param.b32 %r2, [spill_twice_param_3];
+; CHECK-NEXT: ld.param.b32 %r3, [spill_twice_param_4];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd10, %rd11}, [%rd1];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd12, %rd13}, [%rd1+16];
+; CHECK-NEXT: mul.wide.u32 %rd14, %r1, 8;
+; CHECK-NEXT: and.b64 %rd15, %rd14, 56;
+; CHECK-NEXT: add.u64 %rd16, %SP, 0;
+; CHECK-NEXT: add.s64 %rd17, %rd16, %rd15;
+; CHECK-NEXT: shl.b32 %r4, %r1, 3;
+; CHECK-NEXT: cvt.u64.u32 %rd18, %r4;
+; CHECK-NEXT: and.b64 %rd19, %rd18, 56;
+; CHECK-NEXT: add.u64 %rd20, %SP, 64;
+; CHECK-NEXT: add.s64 %rd21, %rd20, %rd19;
+; CHECK-NEXT: mul.wide.u32 %rd22, %r2, 8;
+; CHECK-NEXT: and.b64 %rd23, %rd22, 56;
+; CHECK-NEXT: add.s64 %rd24, %rd16, %rd23;
+; CHECK-NEXT: shl.b32 %r5, %r2, 3;
+; CHECK-NEXT: cvt.u64.u32 %rd25, %r5;
+; CHECK-NEXT: and.b64 %rd26, %rd25, 56;
+; CHECK-NEXT: add.s64 %rd27, %rd20, %rd26;
+; CHECK-NEXT: st.b64 [%SP+120], %rd9;
+; CHECK-NEXT: st.b64 [%SP+112], %rd8;
+; CHECK-NEXT: st.b64 [%SP+104], %rd7;
+; CHECK-NEXT: st.b64 [%SP+96], %rd6;
+; CHECK-NEXT: st.b64 [%SP+88], %rd5;
+; CHECK-NEXT: st.b64 [%SP+80], %rd4;
+; CHECK-NEXT: st.b64 [%SP+72], %rd3;
+; CHECK-NEXT: st.b64 [%SP+64], %rd2;
+; CHECK-NEXT: st.b64 [%rd21], %rd10;
+; CHECK-NEXT: st.b64 [%rd27], %rd11;
+; CHECK-NEXT: shl.b32 %r6, %r3, 3;
+; CHECK-NEXT: cvt.u64.u32 %rd28, %r6;
+; CHECK-NEXT: and.b64 %rd29, %rd28, 56;
+; CHECK-NEXT: add.s64 %rd30, %rd20, %rd29;
+; CHECK-NEXT: st.b64 [%rd30], %rd12;
+; CHECK-NEXT: mul.wide.u32 %rd31, %r3, 8;
+; CHECK-NEXT: and.b64 %rd32, %rd31, 56;
+; CHECK-NEXT: add.s64 %rd33, %rd16, %rd32;
+; CHECK-NEXT: ld.b64 %rd34, [%SP+72];
+; CHECK-NEXT: ld.b64 %rd35, [%SP+64];
+; CHECK-NEXT: ld.b64 %rd36, [%SP+88];
+; CHECK-NEXT: ld.b64 %rd37, [%SP+80];
+; CHECK-NEXT: ld.b64 %rd38, [%SP+104];
+; CHECK-NEXT: ld.b64 %rd39, [%SP+96];
+; CHECK-NEXT: ld.b64 %rd40, [%SP+120];
+; CHECK-NEXT: ld.b64 %rd41, [%SP+112];
+; CHECK-NEXT: st.b64 [%SP+56], %rd9;
+; CHECK-NEXT: st.b64 [%SP+48], %rd8;
+; CHECK-NEXT: st.b64 [%SP+40], %rd7;
+; CHECK-NEXT: st.b64 [%SP+32], %rd6;
+; CHECK-NEXT: st.b64 [%SP+24], %rd5;
+; CHECK-NEXT: st.b64 [%SP+16], %rd4;
+; CHECK-NEXT: st.b64 [%SP+8], %rd3;
+; CHECK-NEXT: st.b64 [%SP], %rd2;
+; CHECK-NEXT: st.b64 [%rd17], %rd10;
+; CHECK-NEXT: st.b64 [%rd24], %rd11;
+; CHECK-NEXT: st.b64 [%rd33], %rd12;
+; CHECK-NEXT: ld.param.b32 %rd42, [spill_twice_param_5];
+; CHECK-NEXT: shl.b64 %rd43, %rd42, 3;
+; CHECK-NEXT: and.b64 %rd44, %rd43, 56;
+; CHECK-NEXT: add.s64 %rd45, %rd16, %rd44;
+; CHECK-NEXT: st.b64 [%rd45], %rd13;
+; CHECK-NEXT: ld.b64 %rd46, [%SP+8];
+; CHECK-NEXT: ld.b64 %rd47, [%SP];
+; CHECK-NEXT: ld.b64 %rd48, [%SP+24];
+; CHECK-NEXT: ld.b64 %rd49, [%SP+16];
+; CHECK-NEXT: ld.b64 %rd50, [%SP+40];
+; CHECK-NEXT: ld.b64 %rd51, [%SP+32];
+; CHECK-NEXT: ld.b64 %rd52, [%SP+56];
+; CHECK-NEXT: ld.b64 %rd53, [%SP+48];
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1072], {%rd53, %rd52};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1056], {%rd51, %rd50};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1040], {%rd49, %rd48};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1024], {%rd47, %rd46};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1144], {%rd41, %rd40};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1128], {%rd39, %rd38};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1112], {%rd37, %rd36};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1096], {%rd35, %rd34};
+; CHECK-NEXT: ret;
+entry:
+ %offset.0 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 0
+ %element.0 = load double, ptr addrspace(3) %offset.0, align 64
+ %offset.1 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 8
+ %element.1 = load double, ptr addrspace(3) %offset.1, align 8
+ %offset.2 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 16
+ %element.2 = load double, ptr addrspace(3) %offset.2, align 8
+ %offset.3 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 24
+ %element.3 = load double, ptr addrspace(3) %offset.3, align 8
+
+; COM: begin chain 1
+ %vector.build0 = insertelement <8 x double> %vector, double %element.0, i32 %idx0
+ %vector.build1 = insertelement <8 x double> %vector.build0, double %element.1, i32 %idx1
+
+; COM: interleave a second chain of insertelements
+ %vector.build1-2 = insertelement <8 x double> %vector.build1, double %element.2, i32 %idx2
+
+; COM: continue chain 1
+ %vector.build2 = insertelement <8 x double> %vector.build1, double %element.2, i32 %idx2
+ %vector.build3 = insertelement <8 x double> %vector.build2, double %element.3, i32 %idx3
+
+; COM: save chain 1
+ %location = getelementptr i8, ptr addrspace(3) %shared.mem, i32 1024
+ store <8 x double> %vector.build3, ptr addrspace(3) %location, align 64
+
+; COM: save chain 2
+ %location-2 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 1096
+ store <8 x double> %vector.build1-2, ptr addrspace(3) %location-2, align 64
+ ret void
+}
|
@llvm/pr-subscribers-llvm-selectiondag Author: Princeton Ferro (Prince781) ChangesA chain of dynamic insertelts (that is: insertelt (insertelt (...)) with dynamic indices) can be spilled at once. This avoids each insertelt being spilled in DAGTypeLegalizer which reduces code size and compile time. Full diff: https://github.com/llvm/llvm-project/pull/162368.diff 2 Files Affected:
diff --git a/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp b/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
index 309f1bea8b77c..3071aac3a511a 100644
--- a/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
+++ b/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
@@ -23445,6 +23445,57 @@ SDValue DAGCombiner::visitINSERT_VECTOR_ELT(SDNode *N) {
// inselt undef, InVal, EltNo --> build_vector < InVal, InVal, ... >
if (InVec.isUndef() && TLI.shouldSplatInsEltVarIndex(VT))
return DAG.getSplat(VT, DL, InVal);
+
+ // For dynamic insertelts, the type legalizer will spill the entire vector.
+ // For a chain of dynamic insertelts, this can be really inefficient and
+ // bad for compile time. If each insertelt is only fed into the next, the
+ // vector is write-only across this chain, and we can just spill once.
+ SmallVector<SDNode *> Seq{N};
+ while (true) {
+ SDValue InVec = Seq.back()->getOperand(0);
+ SDValue EltNo = InVec.getOperand(2);
+ if (!(InVec.getOpcode() == ISD::INSERT_VECTOR_ELT &&
+ !isa<ConstantSDNode>(EltNo)))
+ break;
+ Seq.push_back(InVec.getNode());
+ }
+
+ // Only care about chains, otherwise this instruction can be handled by
+ // the type legalizer just fine.
+ if (Seq.size() > 1) {
+ // In cases where the vector is illegal it will be broken down into parts
+ // and stored in parts - we should use the alignment for the smallest part.
+ Align SmallestAlign = DAG.getReducedAlign(VT, /*UseABI=*/false);
+ SDValue StackPtr = DAG.CreateStackTemporary(VT.getStoreSize(), SmallestAlign);
+ auto &MF = DAG.getMachineFunction();
+ auto FrameIndex = cast<FrameIndexSDNode>(StackPtr.getNode())->getIndex();
+ auto PtrInfo = MachinePointerInfo::getFixedStack(MF, FrameIndex);
+
+ // Begin spilling
+ SDValue InVec = Seq.back()->getOperand(0);
+ SDValue Store = DAG.getStore(DAG.getEntryNode(), DL, InVec, StackPtr, PtrInfo,
+ SmallestAlign);
+
+ // Lower each dynamic insertelt to a store
+ for (SDNode *N : reverse(Seq)) {
+ SDValue Elmnt = N->getOperand(1);
+ SDValue Index = N->getOperand(2);
+
+ // Store the new element. This may be larger than the vector element type,
+ // so use a truncating store.
+ SDValue EltPtr = TLI.getVectorElementPointer(DAG, StackPtr, VT, Index);
+ EVT EltVT = Elmnt.getValueType();
+ Store = DAG.getTruncStore(
+ Store, DL, Elmnt, EltPtr, MachinePointerInfo::getUnknownStack(MF),
+ EltVT, commonAlignment(SmallestAlign,
+ EltVT.getFixedSizeInBits() / 8));
+ }
+
+ // Load the spilled vector
+ SDValue Load = DAG.getLoad(VT, DL, Store, StackPtr, PtrInfo, SmallestAlign);
+ return Load.getValue(0);
+ }
+
return SDValue();
}
diff --git a/llvm/test/CodeGen/NVPTX/vector-spill.ll b/llvm/test/CodeGen/NVPTX/vector-spill.ll
new file mode 100644
index 0000000000000..0a26069b02cc2
--- /dev/null
+++ b/llvm/test/CodeGen/NVPTX/vector-spill.ll
@@ -0,0 +1,215 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 6
+; RUN: llc < %s -mcpu=sm_20 | FileCheck %s
+; RUN: %if ptxas %{ llc < %s -mcpu=sm_20 | %ptxas-verify %}
+target datalayout = "e-i64:64-v16:16-v32:32-n16:32:64"
+target triple = "nvptx64-nvidia-cuda"
+
+; COM: Spill the vector once.
+define ptx_kernel void @spill_once(ptr addrspace(3) %shared.mem, <8 x double> %vector, i32 %idx0, i32 %idx1, i32 %idx2, i32 %idx3) local_unnamed_addr {
+; CHECK-LABEL: spill_once(
+; CHECK: {
+; CHECK-NEXT: .local .align 8 .b8 __local_depot0[64];
+; CHECK-NEXT: .reg .b64 %SP;
+; CHECK-NEXT: .reg .b64 %SPL;
+; CHECK-NEXT: .reg .b64 %rd<39>;
+; CHECK-EMPTY:
+; CHECK-NEXT: // %bb.0: // %entry
+; CHECK-NEXT: mov.b64 %SPL, __local_depot0;
+; CHECK-NEXT: cvta.local.u64 %SP, %SPL;
+; CHECK-NEXT: ld.param.b64 %rd1, [spill_once_param_0];
+; CHECK-NEXT: ld.param.v2.b64 {%rd2, %rd3}, [spill_once_param_1];
+; CHECK-NEXT: ld.param.v2.b64 {%rd4, %rd5}, [spill_once_param_1+16];
+; CHECK-NEXT: ld.param.v2.b64 {%rd6, %rd7}, [spill_once_param_1+32];
+; CHECK-NEXT: ld.param.v2.b64 {%rd8, %rd9}, [spill_once_param_1+48];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd10, %rd11}, [%rd1];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd12, %rd13}, [%rd1+16];
+; CHECK-NEXT: ld.param.b32 %rd14, [spill_once_param_2];
+; CHECK-NEXT: shl.b64 %rd15, %rd14, 3;
+; CHECK-NEXT: and.b64 %rd16, %rd15, 56;
+; CHECK-NEXT: add.u64 %rd17, %SP, 0;
+; CHECK-NEXT: add.s64 %rd18, %rd17, %rd16;
+; CHECK-NEXT: ld.param.b32 %rd19, [spill_once_param_3];
+; CHECK-NEXT: shl.b64 %rd20, %rd19, 3;
+; CHECK-NEXT: and.b64 %rd21, %rd20, 56;
+; CHECK-NEXT: add.s64 %rd22, %rd17, %rd21;
+; CHECK-NEXT: ld.param.b32 %rd23, [spill_once_param_4];
+; CHECK-NEXT: shl.b64 %rd24, %rd23, 3;
+; CHECK-NEXT: and.b64 %rd25, %rd24, 56;
+; CHECK-NEXT: add.s64 %rd26, %rd17, %rd25;
+; CHECK-NEXT: st.b64 [%SP+56], %rd9;
+; CHECK-NEXT: st.b64 [%SP+48], %rd8;
+; CHECK-NEXT: st.b64 [%SP+40], %rd7;
+; CHECK-NEXT: st.b64 [%SP+32], %rd6;
+; CHECK-NEXT: st.b64 [%SP+24], %rd5;
+; CHECK-NEXT: st.b64 [%SP+16], %rd4;
+; CHECK-NEXT: st.b64 [%SP+8], %rd3;
+; CHECK-NEXT: st.b64 [%SP], %rd2;
+; CHECK-NEXT: st.b64 [%rd18], %rd10;
+; CHECK-NEXT: st.b64 [%rd22], %rd11;
+; CHECK-NEXT: st.b64 [%rd26], %rd12;
+; CHECK-NEXT: ld.param.b32 %rd27, [spill_once_param_5];
+; CHECK-NEXT: shl.b64 %rd28, %rd27, 3;
+; CHECK-NEXT: and.b64 %rd29, %rd28, 56;
+; CHECK-NEXT: add.s64 %rd30, %rd17, %rd29;
+; CHECK-NEXT: st.b64 [%rd30], %rd13;
+; CHECK-NEXT: ld.b64 %rd31, [%SP+8];
+; CHECK-NEXT: ld.b64 %rd32, [%SP];
+; CHECK-NEXT: ld.b64 %rd33, [%SP+24];
+; CHECK-NEXT: ld.b64 %rd34, [%SP+16];
+; CHECK-NEXT: ld.b64 %rd35, [%SP+40];
+; CHECK-NEXT: ld.b64 %rd36, [%SP+32];
+; CHECK-NEXT: ld.b64 %rd37, [%SP+56];
+; CHECK-NEXT: ld.b64 %rd38, [%SP+48];
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1072], {%rd38, %rd37};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1056], {%rd36, %rd35};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1040], {%rd34, %rd33};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1024], {%rd32, %rd31};
+; CHECK-NEXT: ret;
+entry:
+ %offset.0 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 0
+ %element.0 = load double, ptr addrspace(3) %offset.0, align 64
+ %offset.1 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 8
+ %element.1 = load double, ptr addrspace(3) %offset.1, align 8
+ %offset.2 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 16
+ %element.2 = load double, ptr addrspace(3) %offset.2, align 8
+ %offset.3 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 24
+ %element.3 = load double, ptr addrspace(3) %offset.3, align 8
+ %vector.build0 = insertelement <8 x double> %vector, double %element.0, i32 %idx0
+ %vector.build1 = insertelement <8 x double> %vector.build0, double %element.1, i32 %idx1
+ %vector.build2 = insertelement <8 x double> %vector.build1, double %element.2, i32 %idx2
+ %vector.build3 = insertelement <8 x double> %vector.build2, double %element.3, i32 %idx3
+ %location = getelementptr i8, ptr addrspace(3) %shared.mem, i32 1024
+ store <8 x double> %vector.build3, ptr addrspace(3) %location, align 64
+ ret void
+}
+
+; COM: Spill the vector twice. Because these are in two different slots, the
+; resulting spill codes may be non-overlapping even though the insertelt
+; sequences overlap.
+define ptx_kernel void @spill_twice(ptr addrspace(3) %shared.mem, <8 x double> %vector, i32 %idx0, i32 %idx1, i32 %idx2, i32 %idx3) local_unnamed_addr {
+; CHECK-LABEL: spill_twice(
+; CHECK: {
+; CHECK-NEXT: .local .align 8 .b8 __local_depot1[128];
+; CHECK-NEXT: .reg .b64 %SP;
+; CHECK-NEXT: .reg .b64 %SPL;
+; CHECK-NEXT: .reg .b32 %r<7>;
+; CHECK-NEXT: .reg .b64 %rd<54>;
+; CHECK-EMPTY:
+; CHECK-NEXT: // %bb.0: // %entry
+; CHECK-NEXT: mov.b64 %SPL, __local_depot1;
+; CHECK-NEXT: cvta.local.u64 %SP, %SPL;
+; CHECK-NEXT: ld.param.b64 %rd1, [spill_twice_param_0];
+; CHECK-NEXT: ld.param.v2.b64 {%rd2, %rd3}, [spill_twice_param_1];
+; CHECK-NEXT: ld.param.v2.b64 {%rd4, %rd5}, [spill_twice_param_1+16];
+; CHECK-NEXT: ld.param.v2.b64 {%rd6, %rd7}, [spill_twice_param_1+32];
+; CHECK-NEXT: ld.param.v2.b64 {%rd8, %rd9}, [spill_twice_param_1+48];
+; CHECK-NEXT: ld.param.b32 %r1, [spill_twice_param_2];
+; CHECK-NEXT: ld.param.b32 %r2, [spill_twice_param_3];
+; CHECK-NEXT: ld.param.b32 %r3, [spill_twice_param_4];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd10, %rd11}, [%rd1];
+; CHECK-NEXT: ld.shared.v2.b64 {%rd12, %rd13}, [%rd1+16];
+; CHECK-NEXT: mul.wide.u32 %rd14, %r1, 8;
+; CHECK-NEXT: and.b64 %rd15, %rd14, 56;
+; CHECK-NEXT: add.u64 %rd16, %SP, 0;
+; CHECK-NEXT: add.s64 %rd17, %rd16, %rd15;
+; CHECK-NEXT: shl.b32 %r4, %r1, 3;
+; CHECK-NEXT: cvt.u64.u32 %rd18, %r4;
+; CHECK-NEXT: and.b64 %rd19, %rd18, 56;
+; CHECK-NEXT: add.u64 %rd20, %SP, 64;
+; CHECK-NEXT: add.s64 %rd21, %rd20, %rd19;
+; CHECK-NEXT: mul.wide.u32 %rd22, %r2, 8;
+; CHECK-NEXT: and.b64 %rd23, %rd22, 56;
+; CHECK-NEXT: add.s64 %rd24, %rd16, %rd23;
+; CHECK-NEXT: shl.b32 %r5, %r2, 3;
+; CHECK-NEXT: cvt.u64.u32 %rd25, %r5;
+; CHECK-NEXT: and.b64 %rd26, %rd25, 56;
+; CHECK-NEXT: add.s64 %rd27, %rd20, %rd26;
+; CHECK-NEXT: st.b64 [%SP+120], %rd9;
+; CHECK-NEXT: st.b64 [%SP+112], %rd8;
+; CHECK-NEXT: st.b64 [%SP+104], %rd7;
+; CHECK-NEXT: st.b64 [%SP+96], %rd6;
+; CHECK-NEXT: st.b64 [%SP+88], %rd5;
+; CHECK-NEXT: st.b64 [%SP+80], %rd4;
+; CHECK-NEXT: st.b64 [%SP+72], %rd3;
+; CHECK-NEXT: st.b64 [%SP+64], %rd2;
+; CHECK-NEXT: st.b64 [%rd21], %rd10;
+; CHECK-NEXT: st.b64 [%rd27], %rd11;
+; CHECK-NEXT: shl.b32 %r6, %r3, 3;
+; CHECK-NEXT: cvt.u64.u32 %rd28, %r6;
+; CHECK-NEXT: and.b64 %rd29, %rd28, 56;
+; CHECK-NEXT: add.s64 %rd30, %rd20, %rd29;
+; CHECK-NEXT: st.b64 [%rd30], %rd12;
+; CHECK-NEXT: mul.wide.u32 %rd31, %r3, 8;
+; CHECK-NEXT: and.b64 %rd32, %rd31, 56;
+; CHECK-NEXT: add.s64 %rd33, %rd16, %rd32;
+; CHECK-NEXT: ld.b64 %rd34, [%SP+72];
+; CHECK-NEXT: ld.b64 %rd35, [%SP+64];
+; CHECK-NEXT: ld.b64 %rd36, [%SP+88];
+; CHECK-NEXT: ld.b64 %rd37, [%SP+80];
+; CHECK-NEXT: ld.b64 %rd38, [%SP+104];
+; CHECK-NEXT: ld.b64 %rd39, [%SP+96];
+; CHECK-NEXT: ld.b64 %rd40, [%SP+120];
+; CHECK-NEXT: ld.b64 %rd41, [%SP+112];
+; CHECK-NEXT: st.b64 [%SP+56], %rd9;
+; CHECK-NEXT: st.b64 [%SP+48], %rd8;
+; CHECK-NEXT: st.b64 [%SP+40], %rd7;
+; CHECK-NEXT: st.b64 [%SP+32], %rd6;
+; CHECK-NEXT: st.b64 [%SP+24], %rd5;
+; CHECK-NEXT: st.b64 [%SP+16], %rd4;
+; CHECK-NEXT: st.b64 [%SP+8], %rd3;
+; CHECK-NEXT: st.b64 [%SP], %rd2;
+; CHECK-NEXT: st.b64 [%rd17], %rd10;
+; CHECK-NEXT: st.b64 [%rd24], %rd11;
+; CHECK-NEXT: st.b64 [%rd33], %rd12;
+; CHECK-NEXT: ld.param.b32 %rd42, [spill_twice_param_5];
+; CHECK-NEXT: shl.b64 %rd43, %rd42, 3;
+; CHECK-NEXT: and.b64 %rd44, %rd43, 56;
+; CHECK-NEXT: add.s64 %rd45, %rd16, %rd44;
+; CHECK-NEXT: st.b64 [%rd45], %rd13;
+; CHECK-NEXT: ld.b64 %rd46, [%SP+8];
+; CHECK-NEXT: ld.b64 %rd47, [%SP];
+; CHECK-NEXT: ld.b64 %rd48, [%SP+24];
+; CHECK-NEXT: ld.b64 %rd49, [%SP+16];
+; CHECK-NEXT: ld.b64 %rd50, [%SP+40];
+; CHECK-NEXT: ld.b64 %rd51, [%SP+32];
+; CHECK-NEXT: ld.b64 %rd52, [%SP+56];
+; CHECK-NEXT: ld.b64 %rd53, [%SP+48];
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1072], {%rd53, %rd52};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1056], {%rd51, %rd50};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1040], {%rd49, %rd48};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1024], {%rd47, %rd46};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1144], {%rd41, %rd40};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1128], {%rd39, %rd38};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1112], {%rd37, %rd36};
+; CHECK-NEXT: st.shared.v2.b64 [%rd1+1096], {%rd35, %rd34};
+; CHECK-NEXT: ret;
+entry:
+ %offset.0 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 0
+ %element.0 = load double, ptr addrspace(3) %offset.0, align 64
+ %offset.1 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 8
+ %element.1 = load double, ptr addrspace(3) %offset.1, align 8
+ %offset.2 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 16
+ %element.2 = load double, ptr addrspace(3) %offset.2, align 8
+ %offset.3 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 24
+ %element.3 = load double, ptr addrspace(3) %offset.3, align 8
+
+; COM: begin chain 1
+ %vector.build0 = insertelement <8 x double> %vector, double %element.0, i32 %idx0
+ %vector.build1 = insertelement <8 x double> %vector.build0, double %element.1, i32 %idx1
+
+; COM: interleave a second chain of insertelements
+ %vector.build1-2 = insertelement <8 x double> %vector.build1, double %element.2, i32 %idx2
+
+; COM: continue chain 1
+ %vector.build2 = insertelement <8 x double> %vector.build1, double %element.2, i32 %idx2
+ %vector.build3 = insertelement <8 x double> %vector.build2, double %element.3, i32 %idx3
+
+; COM: save chain 1
+ %location = getelementptr i8, ptr addrspace(3) %shared.mem, i32 1024
+ store <8 x double> %vector.build3, ptr addrspace(3) %location, align 64
+
+; COM: save chain 2
+ %location-2 = getelementptr i8, ptr addrspace(3) %shared.mem, i32 1096
+ store <8 x double> %vector.build1-2, ptr addrspace(3) %location-2, align 64
+ ret void
+}
|
✅ With the latest revision this PR passed the C/C++ code formatter. |
This change improves ISel compile time on a vector-heavy kernel from 1 hour to 200s. I've included a reduced test case in NVPTX, although this is a target-independent change. |
7a55cca
to
028c586
Compare
A chain of dynamic insertelts with can be spilled at once. This avoids each insertelt being spilled in DAGTypeLegalizer which reduces code size and compile time.
028c586
to
b67c59f
Compare
ret void | ||
} | ||
|
||
; COM: Spill the vector twice. Because these are in two different slots, the |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This isn't spilling?
// vector. For a chain of dynamic insertelts, this can be really | ||
// inefficient and bad for compile time. If each insertelt is only fed | ||
// into the next, the vector is write-only across this chain, and we can | ||
// just spill once. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This isn't spilling
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What's the correct terminology here? We're creating a copy of the vector temporary on the stack. The comments in DAGTypeLegalizer::SplitVecRes_INSERT_VECTOR_ELT
use the same term:
llvm-project/llvm/lib/CodeGen/SelectionDAG/LegalizeVectorTypes.cpp
Lines 2101 to 2140 in 4a9e8ea
// Spill the vector to the stack. | |
// In cases where the vector is illegal it will be broken down into parts | |
// and stored in parts - we should use the alignment for the smallest part. | |
Align SmallestAlign = DAG.getReducedAlign(VecVT, /*UseABI=*/false); | |
SDValue StackPtr = | |
DAG.CreateStackTemporary(VecVT.getStoreSize(), SmallestAlign); | |
auto &MF = DAG.getMachineFunction(); | |
auto FrameIndex = cast<FrameIndexSDNode>(StackPtr.getNode())->getIndex(); | |
auto PtrInfo = MachinePointerInfo::getFixedStack(MF, FrameIndex); | |
SDValue Store = DAG.getStore(DAG.getEntryNode(), dl, Vec, StackPtr, PtrInfo, | |
SmallestAlign); | |
// Store the new element. This may be larger than the vector element type, | |
// so use a truncating store. | |
SDValue EltPtr = TLI.getVectorElementPointer(DAG, StackPtr, VecVT, Idx); | |
Store = DAG.getTruncStore( | |
Store, dl, Elt, EltPtr, MachinePointerInfo::getUnknownStack(MF), EltVT, | |
commonAlignment(SmallestAlign, | |
EltVT.getFixedSizeInBits() / 8)); | |
EVT LoVT, HiVT; | |
std::tie(LoVT, HiVT) = DAG.GetSplitDestVTs(VecVT); | |
// Load the Lo part from the stack slot. | |
Lo = DAG.getLoad(LoVT, dl, Store, StackPtr, PtrInfo, SmallestAlign); | |
// Increment the pointer to the other part. | |
auto Load = cast<LoadSDNode>(Lo); | |
MachinePointerInfo MPI = Load->getPointerInfo(); | |
IncrementPointer(Load, LoVT, MPI, StackPtr); | |
Hi = DAG.getLoad(HiVT, dl, Store, StackPtr, MPI, SmallestAlign); | |
// If we adjusted the original type, we need to truncate the results. | |
std::tie(LoVT, HiVT) = DAG.GetSplitDestVTs(N->getValueType(0)); | |
if (LoVT != Lo.getValueType()) | |
Lo = DAG.getNode(ISD::TRUNCATE, dl, LoVT, Lo); | |
if (HiVT != Hi.getValueType()) | |
Hi = DAG.getNode(ISD::TRUNCATE, dl, HiVT, Hi); |
cast<FrameIndexSDNode>(StackPtr.getNode())->getIndex(); | ||
auto PtrInfo = MachinePointerInfo::getFixedStack(MF, FrameIndex); | ||
|
||
// Begin spilling |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Not spilling
// Only care about chains, otherwise this instruction can be handled by | ||
// the type legalizer just fine. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can you rephrase this comment? This makes it sounds like it's working around a correctness issue in the legalizer
SDValue StackPtr = | ||
DAG.CreateStackTemporary(VT.getStoreSize(), SmallestAlign); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Dynamic vector inserts are not always lowered to the stack, that just happens to be the default operation action.
Or is the vector type legalizer code in a worse state than the LegalizeDAG handling for this?
SDValue StackPtr = | ||
DAG.CreateStackTemporary(VT.getStoreSize(), SmallestAlign); | ||
auto &MF = DAG.getMachineFunction(); | ||
auto FrameIndex = |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
No auto
while (true) { | ||
SDValue InVec = Seq.back()->getOperand(0); | ||
if (InVec.getOpcode() != ISD::INSERT_VECTOR_ELT || | ||
isa<ConstantSDNode>(InVec.getOperand(2))) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
If you have mixed constant and dynamic indexes, I'd assume you wan't to collect and handle them as well
A chain of dynamic insertelts can be spilled at once. This avoids each insertelt being spilled in DAGTypeLegalizer which reduces code size and compile time.